transaction data
TREASURE: A Transformer-Based Foundation Model for High-Volume Transaction Understanding
Yeh, Chin-Chia Michael, Saini, Uday Singh, Dai, Xin, Fan, Xiran, Jain, Shubham, Fan, Yujie, Sun, Jiarui, Wang, Junpeng, Pan, Menghai, Dou, Yingtong, Chen, Yuzhong, Rakesh, Vineeth, Wang, Liang, Zheng, Yan, Das, Mahashweta
Payment networks form the backbone of modern commerce, generating high volumes of transaction records from daily activities. Properly modeling this data can enable applications such as abnormal behavior detection and consumer-level insights for hyper-personalized experiences, ultimately improving people's lives. In this paper, we present TREASURE, TRansformer Engine As Scalable Universal transaction Representation Encoder, a multipurpose transformer-based foundation model specifically designed for transaction data. The model simultaneously captures both consumer behavior and payment network signals (such as response codes and system flags), providing comprehensive information necessary for applications like accurate recommendation systems and abnormal behavior detection. Verified with industry-grade datasets, TREASURE features three key capabilities: 1) an input module with dedicated sub-modules for static and dynamic attributes, enabling more efficient training and inference; 2) an efficient and effective training paradigm for predicting high-cardinality categorical attributes; and 3) demonstrated effectiveness as both a standalone model that increases abnormal behavior detection performance by 111% over production systems and an embedding provider that enhances recommendation models by 104%. We present key insights from extensive ablation studies, benchmarks against production models, and case studies, highlighting valuable knowledge gained from developing TREASURE.
- North America > United States > California > San Mateo County > Foster City (0.40)
- North America > Canada (0.04)
- North America > Puerto Rico (0.04)
- (2 more...)
- Information Technology (0.93)
- Banking & Finance (0.69)
TransactionGPT
Dou, Yingtong, Jiang, Zhimeng, Zhang, Tianyi, Hu, Mingzhi, Xu, Zhichao, Jain, Shubham, Saini, Uday Singh, Fan, Xiran, Sun, Jiarui, Pan, Menghai, Wang, Junpeng, Dai, Xin, Wang, Liang, Yeh, Chin-Chia Michael, Fan, Yujie, Rakesh, Vineeth, Chen, Huiyuan, Bendre, Mangesh, Zhuang, Zhongfang, Li, Xiaoting, Aboagye, Prince, Lai, Vivian, Xu, Minghua, Yang, Hao, Cai, Yiwei, Das, Mahashweta, Chen, Yuzhong
TGPT is designed to understand and generate transaction trajectories while simultaneously supporting a variety of downstream prediction and classification tasks. We introduce a novel 3D-Transformer architecture specifically tailored for capturing the complex dynamics in payment transaction data. This architecture incorporates design innovations that enhance modality fusion and computational efficiency, while seamlessly enabling joint optimization with downstream objectives. Trained on billion-scale real-world transactions, TGPT significantly improves downstream classification performance against a competitive production model and exhibits advantages over baselines in generating future transactions. We conduct extensive empirical evaluations utilizing a diverse collection of company transaction datasets spanning multiple downstream tasks, thereby enabling a thorough assessment of TGPT's effectiveness and efficiency in comparison to established methodologies. Furthermore, we examine the incorporation of LLM-derived embeddings within TGPT and benchmark its performance against fine-tuned LLMs, demonstrating that TGPT achieves superior predictive accuracy as well as faster training and inference. We anticipate that the architectural innovations and practical guidelines from this work will advance foundation models for transaction-like data and catalyze future research in this emerging field.
- North America > United States > Illinois > Cook County > Chicago (0.04)
- North America > United States > California > San Francisco County > San Francisco (0.04)
- North America > United States > California > Los Angeles County > Los Angeles (0.04)
- Information Technology > Security & Privacy (1.00)
- Banking & Finance (1.00)
- Consumer Products & Services (0.93)
- Transportation > Air (0.68)
Hybrid GCN-GRU Model for Anomaly Detection in Cryptocurrency Transactions
Na, Gyuyeon, Park, Minjung, Cha, Hyeonjeong, Kim, Soyoun, Moon, Sunyoung, Lee, Sua, Choi, Jaeyoung, Lee, Hyemin, Chai, Sangmi
Blockchain transaction networks are complex, with evolving temporal patterns and inter - node relationships. To detect illicit activi - ties, we propose a hybrid GCN - GRU model that captu res both structural and sequential features. Using real Bitcoin transaction data (2020 - 2024), our model achieved 0.9470 Accuracy and 0.9807 AUC - ROC, outperform - ing all baselines.
FinSurvival: A Suite of Large Scale Survival Modeling Tasks from Finance
Green, Aaron, Nie, Zihan, Qin, Hanzhen, Seneviratne, Oshani, Bennett, Kristin P.
Survival modeling predicts the time until an event occurs and is widely used in risk analysis; for example, it's used in medicine to predict the survival of a patient based on censored data. There is a need for large-scale, realistic, and freely available datasets for benchmarking artificial intelligence (AI) survival models. In this paper, we derive a suite of 16 survival modeling tasks from publicly available transaction data generated by lending of cryptocurrencies in Decentralized Finance (DeFi). Each task was constructed using an automated pipeline based on choices of index and outcome events. For example, the model predicts the time from when a user borrows cryptocurrency coins (index event) until their first repayment (outcome event). We formulate a survival benchmark consisting of a suite of 16 survival-time prediction tasks (FinSurvival). We also automatically create 16 corresponding classification problems for each task by thresholding the survival time using the restricted mean survival time. With over 7.5 million records, FinSurvival provides a suite of realistic financial modeling tasks that will spur future AI survival modeling research. Our evaluation indicated that these are challenging tasks that are not well addressed by existing methods. FinSurvival enables the evaluation of AI survival models applicable to traditional finance, industry, medicine, and commerce, which is currently hindered by the lack of large public datasets. Our benchmark demonstrates how AI models could assess opportunities and risks in DeFi. In the future, the FinSurvival benchmark pipeline can be used to create new benchmarks by incorporating more DeFi transactions and protocols as the use of cryptocurrency grows.
- Europe > Austria > Vienna (0.14)
- North America > United States > New York > Rensselaer County > Troy (0.04)
- Europe > Monaco (0.04)
- Research Report > New Finding (0.93)
- Research Report > Experimental Study (0.68)
- Health & Medicine > Therapeutic Area > Oncology (1.00)
- Banking & Finance > Trading (1.00)
FinMaster: A Holistic Benchmark for Mastering Full-Pipeline Financial Workflows with LLMs
Jiang, Junzhe, Yang, Chang, Cui, Aixin, Jin, Sihan, Wang, Ruiyu, Li, Bo, Huang, Xiao, Sun, Dongning, Wang, Xinrun
Financial tasks are pivotal to global economic stability; however, their execution faces challenges including labor intensive processes, low error tolerance, data fragmentation, and tool limitations. Although large language models (LLMs) have succeeded in various natural language processing tasks and have shown potential in automating workflows through reasoning and contextual understanding, current benchmarks for evaluating LLMs in finance lack sufficient domain-specific data, have simplistic task design, and incomplete evaluation frameworks. To address these gaps, this article presents FinMaster, a comprehensive financial benchmark designed to systematically assess the capabilities of LLM in financial literacy, accounting, auditing, and consulting. Specifically, FinMaster comprises three main modules: i) FinSim, which builds simulators that generate synthetic, privacy-compliant financial data for companies to replicate market dynamics; ii) FinSuite, which provides tasks in core financial domains, spanning 183 tasks of various types and difficulty levels; and iii) FinEval, which develops a unified interface for evaluation. Extensive experiments over state-of-the-art LLMs reveal critical capability gaps in financial reasoning, with accuracy dropping from over 90% on basic tasks to merely 40% on complex scenarios requiring multi-step reasoning. This degradation exhibits the propagation of computational errors, where single-metric calculations initially demonstrating 58% accuracy decreased to 37% in multimetric scenarios. To the best of our knowledge, FinMaster is the first benchmark that covers full-pipeline financial workflows with challenging tasks. We hope that FinMaster can bridge the gap between research and industry practitioners, driving the adoption of LLMs in real-world financial practices to enhance efficiency and accuracy.
- Asia > China > Hong Kong (0.04)
- North America > Mexico (0.04)
- Asia > Singapore (0.04)
- (2 more...)
- Financial News (1.00)
- Workflow (0.90)
Synthetic Data for Robust AI Model Development in Regulated Enterprises
In today's business landscape, organizations need to find the right balance between using their customers' data ethically to power AI solutions and being compliant regarding data privacy and data usage regulations. In this paper, we discuss synthetic data as a possible solution to this dilemma. Synthetic data is simulated data that mimics the real data. We explore how organizations in heavily regulated industries, such as financial institutions or healthcare organizations, can leverage synthetic data to build robust AI solutions while staying compliant. We demonstrate that synthetic data offers two significant advantages by allowing AI models to learn from more diverse data and by helping organizations stay compliant against data privacy laws with the use of synthetic data instead of customer information. We discuss case studies to show how synthetic data can be effectively used in the finance and healthcare sector while discussing the challenges of using synthetic data and some ethical questions it raises. Our research finds that synthetic data could be a game-changer for AI in regulated industries. The potential can be realized when industry, academia, and regulators collaborate to build solutions. We aim to initiate discussions on the use of synthetic data to build ethical, responsible, and effective AI systems in regulated enterprise industries.
- North America > United States > Washington > King County > Bellevue (0.04)
- Asia > Middle East > Oman > Muscat Governorate > Muscat (0.04)
- Law (1.00)
- Information Technology > Security & Privacy (1.00)
- Health & Medicine (1.00)
- Banking & Finance (1.00)
Instruction-Based Fine-tuning of Open-Source LLMs for Predicting Customer Purchase Behaviors
Ergul, Halil Ibrahim, Balcisoy, Selim, Bozkaya, Burcin
In this study, the performance of various predictive models, including probabilistic baseline, CNN, LSTM, and finetuned LLMs, in forecasting merchant categories from financial transaction data have been evaluated. Utilizing datasets from Bank A for training and Bank B for testing, the superior predictive capabilities of the fine-tuned Mistral Instruct model, which was trained using customer data converted into natural language format have been demonstrated. The methodology of this study involves instruction fine-tuning Mistral via LoRA (LowRank Adaptation of Large Language Models) to adapt its vast pre-trained knowledge to the specific domain of financial transactions. The Mistral model significantly outperforms traditional sequential models, achieving higher F1 scores in the three key merchant categories of bank transaction data (grocery, clothing, and gas stations) that is crucial for targeted marketing campaigns. This performance is attributed to the model's enhanced semantic understanding and adaptability which enables it to better manage minority classes and predict transaction categories with greater accuracy. These findings highlight the potential of LLMs in predicting human behavior.
- Europe > Middle East > Republic of Türkiye > Istanbul Province > Istanbul (0.04)
- Asia > Middle East > Republic of Türkiye > Istanbul Province > Istanbul (0.04)
- Europe > Slovenia > Drava > Municipality of Benedikt > Benedikt (0.04)
- Banking & Finance (1.00)
- Transportation > Ground > Road (0.35)
Dynamic Feature Fusion: Combining Global Graph Structures and Local Semantics for Blockchain Fraud Detection
Sheng, Zhang, Song, Liangliang, Wang, Yanbin
--The advent of blockchain technology has facilitated the widespread adoption of smart contracts in the financial sector . However, current fraud detection methodologies exhibit limitations in capturing both global structural patterns within transaction networks and local semantic relationships embedded in transaction data. Most existing models focus on either structural information or semantic features individually, leading to suboptimal performance in detecting complex fraud patterns.In this paper, we propose a dynamic feature fusion model that combines graph-based representation learning and semantic feature extraction for blockchain fraud detection. Specifically, we construct global graph representations to model account relationships and extract local contextual features from transaction data. A dynamic multimodal fusion mechanism is introduced to adaptively integrate these features, enabling the model to capture both structural and semantic fraud patterns effectively. We further develop a comprehensive data processing pipeline, including graph construction, temporal feature enhancement, and text preprocessing. Experimental results on large-scale real-world blockchain datasets demonstrate that our method outperforms existing benchmarks across accuracy, F1 score, and recall metrics. This work highlights the importance of integrating structural relationships and semantic similarities for robust fraud detection and offers a scalable solution for securing blockchain systems. LOCKCHAIN technology has developed rapidly in recent years and has triggered far-reaching changes in several fields, especially in the financial industry [1]. However, as the popularity of blockchain applications grows, so does the significant increase in fraudulent behaviors it has brought about, with serious implications for society [2]. Blockchain technology, due to its decentralization and transparency, has become a tool for unscrupulous individuals to exploit, although it provides greater security and efficiency in financial transactions [3]. For example, the application of blockchain technology in the supply chain is seen as an effective means to enhance transparency and traceability, but it also faces a crisis of social trust due to fraudulent behavior [4].
- Law Enforcement & Public Safety > Fraud (1.00)
- Banking & Finance (1.00)
- Information Technology > e-Commerce > Financial Technology (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Text Processing (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.94)
Generative AI for Banks: Benchmarks and Algorithms for Synthetic Financial Transaction Data
Karst, Fabian Sven, Chong, Sook-Yee, Antenor, Abigail A., Lin, Enyu, Li, Mahei Manhai, Leimeister, Jan Marco
The banking sector, as a data-driven industry, relies on the availability of high-quality data to create value and protect its customers. The synergy between recent deep learning (DL) advancements, and the sector's data needs presents a growth potential of USD$4.6 trillion by 2035 (Accenture, 2017). However, deploying DL models is challenging due to the need for large, high-quality training data (Ryll et al., 2020), a difficulty made worse by the intricacy of financial transaction data (with complex data patterns and time-related characteristics), and strict regulations that limit data sharing (EU Regulation 2016/679, PCI DSS v4.0). One possible solution is to use synthetic data which is artificially generated rather than drawn from real-world events to increase samples in the minority class (Jordon et al., 2022), and allow safe data sharing between financial institutions while protecting privacy (Karst et al., 2024). This approach is essential for improving models used in assessing risks and detecting fraud.
- Information Technology > Security & Privacy (1.00)
- Banking & Finance (1.00)
Blockchain Data Analysis in the Era of Large-Language Models
Toyoda, Kentaroh, Wang, Xiao, Li, Mingzhe, Gao, Bo, Wang, Yuan, Wei, Qingsong
Blockchain data analysis is essential for deriving insights, tracking transactions, identifying patterns, and ensuring the integrity and security of decentralized networks. It plays a key role in various areas, such as fraud detection, regulatory compliance, smart contract auditing, and decentralized finance (DeFi) risk management. However, existing blockchain data analysis tools face challenges, including data scarcity, the lack of generalizability, and the lack of reasoning capability. We believe large language models (LLMs) can mitigate these challenges; however, we have not seen papers discussing LLM integration in blockchain data analysis in a comprehensive and systematic way. This paper systematically explores potential techniques and design patterns in LLM-integrated blockchain data analysis. We also outline prospective research opportunities and challenges, emphasizing the need for further exploration in this promising field. This paper aims to benefit a diverse audience spanning academia, industry, and policy-making, offering valuable insights into the integration of LLMs in blockchain data analysis.
- Information Technology > Security & Privacy (1.00)
- Banking & Finance > Trading (1.00)
- Law Enforcement & Public Safety > Fraud (0.90)
- Information Technology > e-Commerce > Financial Technology (1.00)
- Information Technology > Data Science (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)